14 research outputs found

    Leveraging image noise: source camera identification and increased robustness of convolutional neural networks

    Get PDF
    Image noise can be a boon or a bane depending on the application. For instance, it is a boon in digital image forensics as it helps law enforcement agencies to identify the source camera device from an image. This becomes crucial when dealing with, for example, digital content of child sexual abuse to identify the offender. In this work, methods were developed to extract and match forensic traces from two images using a Convolutional Neural Network (ConvNet). Furthermore, it was observed that homogeneous regions in an image contain forensic traces (in the form of sensor noise) that are least affected by the scene content. This idea was leveraged to develop a camera identification method using ConvNets that resulted in the state-of-the-art camera model identification accuracy of 99.01% on the Dresden dataset. A study of videos in the Vision dataset shows that the methods developed for images do not plainly translate to videos. A category of video frames known as I-frames was more suitable due to the least compression. Proposed methods identified source cameras even under video compression.Image noise can break a ConvNet when it is presented with noisy out-of-distribution images. Two techniques are presented to address this issue. One is a CORF-based pre-processing step, which emphasizes the high-level contours in an image while suppressing the low-level texture and noise. The other method, again biologically inspired, makes architectural changes to a ConvNet with PushPull-Conv. This makes models robust to high-frequency corruptions with a little compromise on clean test data

    On Improving Generalization of CNN-Based Image Classification with Delineation Maps Using the CORF Push-Pull Inhibition Operator

    Get PDF
    Deployed image classification pipelines are typically dependent on the images captured in real-world environments. This means that images might be affected by different sources of perturbations (e.g. sensor noise in low-light environments). The main challenge arises by the fact that image quality directly impacts the reliability and consistency of classification tasks. This challenge has, hence, attracted wide interest within the computer vision communities. We propose a transformation step that attempts to enhance the generalization ability of CNN models in the presence of unseen noise in the test set. Concretely, the delineation maps of given images are determined using the CORF push-pull inhibition operator. Such an operation transforms an input image into a space that is more robust to noise before being processed by a CNN. We evaluated our approach on the Fashion MNIST data set with an AlexNet model. It turned out that the proposed CORF-augmented pipeline achieved comparable results on noise-free images to those of a conventional AlexNet classification model without CORF delineation maps, but it consistently achieved significantly superior performance on test images perturbed with different levels of Gaussian and uniform noise

    Device-based image matching with similarity learning by convolutional neural networks that exploit the underlying camera sensor pattern noise

    Get PDF
    One of the challenging problems in digital image forensics is the capability to identify images that are captured by the same camera device. This knowledge can help forensic experts in gathering intelligence about suspects by analyzing digital images. In this paper, we propose a two-part network to quantify the likelihood that a given pair of images have the same source camera, and we evaluated it on the benchmark Dresden data set containing 1851 images from 31 different cameras. To the best of our knowledge, we are the first ones addressing the challenge of device-based image matching. Though the proposed approach is not yet forensics ready, our experiments show that this direction is worth pursuing, achieving at this moment 85 percent accuracy. This ongoing work is part of the EU-funded project 4NSEEK concerned with forensics against child sexual abuse.Comment: 7 pages, 4 figures, conference pape

    Video Camera Identification from Sensor Pattern Noise with a Constrained ConvNet

    Get PDF
    The identification of source cameras from videos, though it is a highly relevant forensic analysis topic, has been studied much less than its counterpart that uses images. In this work we propose a method to identify the source camera of a video based on camera specific noise patterns that we extract from video frames. For the extraction of noise pattern features, we propose an extended version of a constrained convolutional layer capable of processing color inputs. Our system is designed to classify individual video frames which are in turn combined by a majority vote to identify the source camera. We evaluated this approach on the benchmark VISION data set consisting of 1539 videos from 28 different cameras. To the best of our knowledge, this is the first work that addresses the challenge of video camera identification on a device level. The experiments show that our approach is very promising, achieving up to 93.1% accuracy while being robust to the WhatsApp and YouTube compression techniques. This work is part of the EU-funded project 4NSEEK focused on forensics against child sexual abuse.Comment: Paper Accepted in - 10th International Conference on Pattern Recognition Applications and Methods (ICPRAM 2021

    Source Camera Device Identification from Videos

    Get PDF
    Source camera identification is an important and challenging problem in digital image forensics. The clues of the device used to capture the digital media are very useful for Law Enforcement Agencies (LEAs), especially to help them collect more intelligence in digital forensics. In our work, we focus on identifying the source camera device based on digital videos using deep learning methods. In particular, we evaluate deep learning models with increasing levels of complexity for source camera identification and show that with such sophistication the scene-suppression techniques do not aid in model performance. In addition, we mention several common machine learning strategies that are counter-productive in achieving a high accuracy for camera identification. We conduct systematic experiments using 28 devices from the VISION data set and evaluate the model performance on various video scenarios—flat (i.e., homogeneous), indoor, and outdoor and evaluate the impact on classification accuracy when the videos are shared via social media platforms such as YouTube and WhatsApp. Unlike traditional PRNU-noise (Photo Response Non-Uniform)-based methods which require flat frames to estimate camera reference pattern noise, the proposed method has no such constraint and we achieve an accuracy of 72.75±1.1%72.75 \pm 1.1 \%on the benchmark VISION data set. Furthermore, we also achieve state-of-the-art accuracy of 71.75%71.75\%on the QUFVD data set in identifying 20 camera devices. These two results are the best ever reported on the VISION and QUFVD data sets. Finally, we demonstrate the runtime efficiency of the proposed approach and its advantages to LEAs

    Camera model identification based on forensic traces extracted from homogeneous patches

    Get PDF
    A crucial challenge in digital image forensics is to identify the source camera model used to generate given images. This is of prime importance, especially for Law Enforcement Agencies in their investigations of Child Sexual Abuse Material found in darknets or seized storage devices. In this work, we address this challenge by proposing a solution that is characterized by two main contributions. It relies on the extraction of rather small homogeneous regions that we extract very efficiently from the integral image, and on a hierarchical classification approach with convolutional neural networks as the underlying models. We rely on homogeneous regions as they contain camera traces that are less distorted than regions with high-level scene content. The hierarchical approach that we propose is important for scaling up and making minimal modifications when new cameras are added. Furthermore, this scheme performs better than the traditional single classifier approach. By means of thorough experimentation on the publicly available Dresden data set, we achieve an accuracy of 99.01% with 5-fold cross-validation on the ‘natural’ subset of this data set. To the best of our knowledge, this is the best result ever reported for Dresden data set
    corecore